Incorporating A Rich Linguistic Model into Whole-Book Recognition
نویسندگان
چکیده
Whole-book recognition, a technique that improves recognition of book-images using fully automatic mutual-entropybased model adaptation, has achieved character error rate as low as 1.9% on 50 pages of real book images in our previous publications. However, the linguistic model for word recognition was simple, assuming a uniform distribution on the words in the dictionary, so that the algorithm is unaware of prior word-occurrence distribution. As a result, the statistics of the output transcript differs largely from that of a real distribution. In this paper, we propose a post-processing technique that improves the existing whole-book recognition results by applying the constraints of a rich linguistic model a prior word-occurrence distribution. This technique further drives the character error rate down from 1.9% to 0.97%. We also show that the wholebook recognition algorithm combined with this post-processing technique shows faster improvements in which word error rates fall monotonically with passage length.
منابع مشابه
Incorporating linguistic post-processing into whole-book recognition
We describe a technique of linguistic post-processing of whole-book recognition results. Whole-book recognition is a technique that improves recognition of book images using fully automatic cross-entropy-based model adaptation. In previous published works, word recognition was performed on individual words separately, without awaring passage-level information such as word-occurrence frequencies...
متن کاملWhole-book recognition using mutual-entropy-driven model adaptation
We describe an approach to unsupervised high-accuracy recognition of the textual contents of an entire book using fully automatic mutual-entropy-based model adaptation. Given images of all the pages of a book together with approximate models of image formation (e.g. a character-image classifier) and linguistics (e.g. a word-occurrence probability model), we detect evidence for disagreements bet...
متن کاملIncorporating linguistic knowledge and automatic baseform generation in acoustic subword unit based speech recognition
A major challenge in speech recognition based on acoustic subword units is creating a lexicon which is robust to interand intra-speaker variations. In this paper we present two di erent approaches for incorporating simple word-level linguistic knowledge into the labelling step of the training procedure. The proposed systems also utilise a scheme for combined optimisation of baseforms and subwor...
متن کاملIncorporating Cognitive Linguistic Insights into Classrooms: the Case of Iranian Learners’ Acquisition of If-Clauses
Cognitive linguistics gives the most inclusive, consistent description of how language is organized, used and learned to date. Cognitive linguistics contains a great number of concepts that are useful to second language learners. If-clauses in English, on the other hand, remain intriguing for foreign language learners to struggle with, due to their intrinsic intricacies. EFL grammar books are ...
متن کاملBeyond N-Grams: Can Linguistic Sophistication Improve Language Modeling?
It seems obvious that a successful model of natural language would incorporate a great deal of both linguistic and world knowledge. Interestingly, state of the art language models for speech recognition are based on a very crude linguistic model, namely conditioning the probability of a word on a small fixed number of preceding words. Despite many attempts to incorporate more sophisticated info...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009